Aalborg
- North America > Canada > Ontario > Toronto (0.04)
- North America > Canada > Alberta (0.04)
- Europe > Denmark > North Jutland > Aalborg (0.04)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis (0.68)
- Information Technology > Artificial Intelligence > Natural Language (0.68)
- North America > Canada > Ontario > Toronto (0.04)
- North America > Canada > Alberta (0.04)
- Europe > Denmark > North Jutland > Aalborg (0.04)
- (4 more...)
- Overview (0.67)
- Research Report (0.46)
- North America > United States > California > Orange County > Irvine (0.14)
- North America > Canada (0.04)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Europe > Denmark > North Jutland > Aalborg (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty (0.93)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (0.69)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.46)
- North America > United States > New York > New York County > New York City (0.14)
- Asia > China > Beijing > Beijing (0.04)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- (21 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.94)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (0.70)
- North America > Canada > Ontario > Toronto (0.04)
- North America > United States (0.04)
- North America > Costa Rica (0.04)
- (2 more...)
- North America > United States (0.14)
- North America > Canada > Quebec (0.04)
- North America > Canada > Ontario (0.04)
- (2 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.93)
- Health & Medicine > Pharmaceuticals & Biotechnology (1.00)
- Government (0.67)
- Asia > China > Beijing > Beijing (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
- (7 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Europe > Netherlands > South Holland > Delft (0.04)
- Europe > Denmark > North Jutland > Aalborg (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Austria (0.04)
Reliable Explanations or Random Noise? A Reliability Metric for XAI
Sengupta, Poushali, Maharjan, Sabita, Eliassen, Frank, Pandey, Shashi Raj, Zhang, Yan
In recent years, explaining decisions made by complex machine learning models has become essential in high-stakes domains such as energy systems, healthcare, finance, and autonomous systems. However, the reliability of these explanations, namely, whether they remain stable and consistent under realistic, non-adversarial changes, remains largely unmeasured. Widely used methods such as SHAP and Integrated Gradients (IG) are well-motivated by axiomatic notions of attribution, yet their explanations can vary substantially even under system-level conditions, including small input perturbations, correlated representations, and minor model updates. Such variability undermines explanation reliability, as reliable explanations should remain consistent across equivalent input representations and small, performance-preserving model changes. We introduce the Explanation Reliability Index (ERI), a family of metrics that quantifies explanation stability under four reliability axioms: robustness to small input perturbations, consistency under feature redundancy, smoothness across model evolution, and resilience to mild distributional shifts. For each axiom, we derive formal guarantees, including Lipschitz-type bounds and temporal stability results. We further propose ERI-T, a dedicated measure of temporal reliability for sequential models, and introduce ERI-Bench, a benchmark designed to systematically stress-test explanation reliability across synthetic and real-world datasets. Experimental results reveal widespread reliability failures in popular explanation methods, showing that explanations can be unstable under realistic deployment conditions. By exposing and quantifying these instabilities, ERI enables principled assessment of explanation reliability and supports more trustworthy explainable AI (XAI) systems.
- North America > Canada > Ontario > Toronto (0.14)
- Europe > Norway > Eastern Norway > Oslo (0.04)
- Europe > Denmark > North Jutland > Aalborg (0.04)
- Research Report > Experimental Study (0.67)
- Research Report > New Finding (0.67)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.92)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.92)
- Information Technology > Artificial Intelligence > Natural Language (0.87)
Improving the Linearized Laplace Approximation via Quadratic Approximations
Jiménez, Pedro, Ortega, Luis A., Morales-Álvarez, Pablo, Hernández-Lobato, Daniel
Deep neural networks (DNNs) often produce overconfident out-of-distribution predictions, motivating Bayesian uncertainty quantification. The Linearized Laplace Approximation (LLA) achieves this by lin-earizing the DNN and applying Laplace inference to the resulting model. Importantly, the linear model is also used for prediction. We argue this linearization in the posterior may degrade fidelity to the true Laplace approximation. To alleviate this problem, without increasing significantly the computational cost, we propose the Quadratic Laplace Approximation (QLA). QLA approximates each second order factor in the approximate Laplace log-posterior using a rank-one factor obtained via efficient power iterations. QLA is expected to yield a posterior precision closer to that of the full Laplace without forming the full Hessian, which is typically intractable. For prediction, QLA also uses the linearized model.
- Europe > Spain > Galicia > Madrid (0.06)
- Europe > Spain > Andalusia > Granada Province > Granada (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- (2 more...)